ALGORITHM OF AMSGrad AND CHAOS OPTIMIZATION IN MULTILAYERED NEURON NETWORKS WITH STOCHASTIC GRADIENT DESCENT
نویسندگان
چکیده
In this paper, the AMSGrad stochastic optimization method was tested using logistic function that describes doubling process and Fourier spectra of error function. The implementation gradient descent algorithm, AMSGrad, done for a multilayer neural network with hidden layers. program recognizing printed digits written Python software environment. array each digit consisted set "0" "1" size 4x7. sample contained 5 possible distortions 3 arrays did not correspond to any digits. analysis influence value hyperparameters beta1 , beta2 learning rate on optimizing teaching network, which layers 28 neurons each, carried out. We constructed branching diagrams based these parameters. found hyperparameter contribution linear function, is associated number local global minima in retraining network. square contribution, block structure formation blocks processes. If alpha greater than rate, there transition chaotic state, accompanied by both multiple passage through minimum and, apparently, appearance minima. At such speed learning, optimizer practically does work, but presence i.e. gradient, general picture chaos described application shown lead better compared conventional even at optimal (the when existing doubles). Keywords: methods, diagrams.
منابع مشابه
Ant Colony Optimization and Stochastic Gradient Descent
In this article, we study the relationship between the two techniques known as ant colony optimization (ACO) and stochastic gradient descent. More precisely, we show that some empirical ACO algorithms approximate stochastic gradient descent in the space of pheromones, and we propose an implementation of stochastic gradient descent that belongs to the family of ACO algorithms. We then use this i...
متن کاملImproved Stochastic gradient descent algorithm for SVM
In order to improve the efficiency and classification ability of Support vector machines (SVM) based on stochastic gradient descent algorithm, three algorithms of improved stochastic gradient descent (SGD) are used to solve support vector machine, which are Momentum, Nesterov accelerated gradient (NAG), RMSprop. The experimental results show that the algorithm based on RMSprop for solving the l...
متن کاملMutiple-gradient Descent Algorithm for Multiobjective Optimization
The steepest-descent method is a well-known and effective single-objective descent algorithm when the gradient of the objective function is known. Here, we propose a particular generalization of this method to multi-objective optimization by considering the concurrent minimization of n smooth criteria {J i } (i = 1,. .. , n). The novel algorithm is based on the following observation: consider a...
متن کاملDistributed Stochastic Optimization via Adaptive Stochastic Gradient Descent
Stochastic convex optimization algorithms are the most popular way to train machine learning models on large-scale data. Scaling up the training process of these models is crucial in many applications, but the most popular algorithm, Stochastic Gradient Descent (SGD), is a serial algorithm that is surprisingly hard to parallelize. In this paper, we propose an efficient distributed stochastic op...
متن کاملStochastic Gradient Descent with GPGPU
We show how to optimize a Support Vector Machine and a predictor for Collaborative Filtering with Stochastic Gradient Descent on the GPU, achieving 1.66 to 6-times accelerations compared to a CPUbased implementation. The reference implementations are the Support Vector Machine by Bottou and the BRISMF predictor from the Netflix Prices winning team. Our main idea is to create a hash function of ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Elektronìka ta ìnformacìjnì tehnologìï
سال: 2023
ISSN: ['2224-0888', '2224-087X']
DOI: https://doi.org/10.30970/eli.21.7